In 2017, when Cloudflare launched Workers, the pitch was almost a curiosity: run small pieces of JavaScript in the same nodes serving CDNs, with minimal latency and no servers to manage. Eight years later, that curiosity has turned into an application platform with database, object storage, queues, pub/sub, AI inference, and enterprise integrations. The question is no longer “what is this for?” but “where does it fit in my architecture?” And that question deserves a nuanced answer.
What it offers today
The basic architecture hasn’t changed: each Worker is a function running in V8 isolates, deployed globally in seconds, cold-starting in under 5 ms, billed per request and CPU ms. That’s still the core of the product and the reason Workers beats Lambda in every case where cold-start latency matters.
What has changed is everything around the Worker. D1 (distributed SQLite with regional replication) has left beta and is usable in production for moderate workloads. R2 (S3-compatible storage with no egress fees) has displaced S3 in many deployments where outbound traffic outweighs storage. Durable Objects, the strangest primitive in the catalog, is today the basis on which collaborative edits, chat rooms, and distributed-state coordinators without explicit servers are built. Queues, Pub/Sub, and Workflows complete the puzzle for event-oriented architectures.
Workers AI is the most recent addition and probably the most commercially relevant. It lets you run a wide selection of open models (Llama, Mistral, Stable Diffusion, embedding models) directly from a Worker, with no GPU infrastructure to manage. Latency is reasonable and the per-token pricing model is more predictable than that of providers billing by GPU-hour.
Where it fits well
The flagship case is still any workload where “close to the user” matters. API gateways, personalization logic, edge A/B testing, header rewriting, route-specific bot protection: Workers is almost unbeatable here. Latency from anywhere on the planet to a nearby Worker is tens of milliseconds, and usage-based billing avoids the classic problem of idle servers in secondary regions.
The second case where it stands out is complete applications that fit in a Worker’s CPU and memory budget (128 MB, configurable CPU time up to 5 minutes on paid plans). For small or mid-sized SaaS, with D1 as the database and R2 as storage, you can cover almost the whole stack without touching other infrastructure. Many niche products are launching this way.
The third case, which has grown a lot over the last year, is running simple AI pipelines. A Worker receives a request, calls Workers AI to generate embeddings or complete a prompt, queries Vectorize (Cloudflare’s vector database), and returns the result. The whole pipeline lives in Cloudflare’s network, with no egress. For light RAG systems, this topology is hard to beat in cost and simplicity.
Where it still doesn’t fit
Despite the growing catalog, there are cases where Workers isn’t the right pick, and it’s worth knowing ahead of time.
Large in-memory state (more than 128 MB) doesn’t fit. You can split across several Workers or lean on Durable Objects, but anyone needing a large, persistent process will struggle.
Libraries in languages other than JavaScript/TypeScript still have limitations. The runtime supports Rust and Python via WASM and Pyodide respectively, but the available library ecosystem is narrower. If your app depends heavily on specific scientific Python or compiled C++ packages, Workers isn’t your spot.
Relational databases with high global consistency requirements are also delicate. D1 works well for mid-sized apps, but eventual consistency across regions means it’s not a direct replacement for a multi-region Postgres with synchronous replication. For banking or similar workloads, you’ll want to look elsewhere.
Pricing and the fine print
Cloudflare has been adjusting its pricing model over the last year, and there are nuances worth looking at. The free plan remains generous for experimenting. Paid plans (Workers Paid, Workers Unbound, now unified) bill by requests and CPU used, with reasonable minimums.
What deserves attention are the “advanced” features with separate prices: Workers AI bills per token or inference, R2 charges for operations even without egress, and D1 charges for rows read/written. All reasonable in isolation, but you need to do the math rather than assume a single flat rate.
The other interesting change is the arrival of an Enterprise plan with dedicated agreements and support, and integration with corporate marketplaces (AWS Marketplace, Azure). This is a clear sign Cloudflare is entering conversations it didn’t access before.
How to think about the decision
The useful question when evaluating Workers isn’t “is it better than Lambda?” but “how much weight does edge logic carry in my architecture?” If the answer is a lot (high-traffic public APIs, authentication logic, region-specific routes, heavy-cache APIs), Workers wins comfortably. If the answer is little (apps with heavy central backend processing and little edge logic), the operational cost of splitting your architecture across two platforms probably doesn’t pay off.
The incremental path that strikes me as most sensible, and that I’ve seen work, is starting with a single Worker in front of your current stack, to handle lightweight auth, cache, or rewrites. From there, move pieces to the edge as it makes technical and economic sense. Avoid redesigning everything from scratch in Workers: the platform is powerful, but its mental model differs from a traditional server, and taking it all on at once tends to end in avoidable friction.
What’s coming
What Cloudflare signaled in its latest Birthday Week suggests the focus for 2025 is two things: deepening Workers AI (more models, better pricing, lower latency) and pushing Durable Objects as a fundamental primitive for stateful apps. If both bets pay off, Workers will become a platform competing head-on with Lambda+DynamoDB for any workload that doesn’t require heavy central processing.
Eight years after its quiet debut, Workers has stopped being a niche tool. It’s still the most elegant option for edge logic, and it’s earned the right to be on the shortlist for complete applications. It’s one of the few platforms where technical innovation and business model have evolved coherently, and that explains why teams that try Workers rarely go back.